65 research outputs found

    Embeddings of Decomposition Spaces into Sobolev and BV Spaces

    Full text link
    In the present paper, we investigate whether an embedding of a decomposition space D(Q,Lp,Y)\mathcal{D}\left(\mathcal{Q},L^{p},Y\right) into a given Sobolev space Wk,q(Rd)W^{k,q}(\mathbb{R}^{d}) exists. As special cases, this includes embeddings into Sobolev spaces of (homogeneous and inhomogeneous) Besov spaces, (α\alpha)-modulation spaces, shearlet smoothness spaces and also of a large class of wavelet coorbit spaces, in particular of shearlet-type coorbit spaces. Precisely, we will show that under extremely mild assumptions on the covering Q=(Qi)iI\mathcal{Q}=\left(Q_{i}\right)_{i\in I}, we have D(Q,Lp,Y)Wk,q(Rd)\mathcal{D}\left(\mathcal{Q},L^{p},Y\right)\hookrightarrow W^{k,q}(\mathbb{R}^{d}) as soon as pqp\leq q and Yu(k,p,q)q(I)Y\hookrightarrow\ell_{u^{\left(k,p,q\right)}}^{q^{\triangledown}}\left(I\right) hold. Here, q=min{q,q}q^{\triangledown}=\min\left\{ q,q'\right\} and the weight u(k,p,q)u^{\left(k,p,q\right)} can be easily computed, only based on the covering Q\mathcal{Q} and on the parameters k,p,qk,p,q. Conversely, a necessary condition for existence of the embedding is that pqp\leq q and Y0(I)u(k,p,q)q(I)Y\cap\ell_{0}\left(I\right)\hookrightarrow\ell_{u^{\left(k,p,q\right)}}^{q}\left(I\right) hold, where 0(I)\ell_{0}\left(I\right) denotes the space of finitely supported sequences on II. All in all, for the range q(0,2]{}q \in (0,2]\cup\{\infty\}, we obtain a complete characterization of existence of the embedding in terms of readily verifiable criteria. We can also completely characterize existence of an embedding of a decomposition space into a BV space

    Optimal approximation of piecewise smooth functions using deep ReLU neural networks

    Full text link
    We study the necessary and sufficient complexity of ReLU neural networks---in terms of depth and number of weights---which is required for approximating classifier functions in L2L^2. As a model class, we consider the set Eβ(Rd)\mathcal{E}^\beta (\mathbb R^d) of possibly discontinuous piecewise CβC^\beta functions f:[1/2,1/2]dRf : [-1/2, 1/2]^d \to \mathbb R, where the different smooth regions of ff are separated by CβC^\beta hypersurfaces. For dimension d2d \geq 2, regularity β>0\beta > 0, and accuracy ε>0\varepsilon > 0, we construct artificial neural networks with ReLU activation function that approximate functions from Eβ(Rd)\mathcal{E}^\beta(\mathbb R^d) up to L2L^2 error of ε\varepsilon. The constructed networks have a fixed number of layers, depending only on dd and β\beta, and they have O(ε2(d1)/β)O(\varepsilon^{-2(d-1)/\beta}) many nonzero weights, which we prove to be optimal. In addition to the optimality in terms of the number of weights, we show that in order to achieve the optimal approximation rate, one needs ReLU networks of a certain depth. Precisely, for piecewise Cβ(Rd)C^\beta(\mathbb R^d) functions, this minimal depth is given---up to a multiplicative constant---by β/d\beta/d. Up to a log factor, our constructed networks match this bound. This partly explains the benefits of depth for ReLU networks by showing that deep networks are necessary to achieve efficient approximation of (piecewise) smooth functions. Finally, we analyze approximation in high-dimensional spaces where the function ff to be approximated can be factorized into a smooth dimension reducing feature map τ\tau and classifier function gg---defined on a low-dimensional feature space---as f=gτf = g \circ \tau. We show that in this case the approximation rate depends only on the dimension of the feature space and not the input dimension.Comment: Generalized some estimates to LpL^p norms for $0<p<\infty

    Wavelet Coorbit Spaces viewed as Decomposition Spaces

    Full text link
    In this paper we show that the Fourier transform induces an isomorphism between the coorbit spaces defined by Feichtinger and Gr\"ochenig of the mixed, weighted Lebesgue spaces Lvp,qL_{v}^{p,q} with respect to the quasi-regular representation of a semi-direct product RdH\mathbb{R}^{d}\rtimes H with suitably chosen dilation group HH, and certain decomposition spaces D(Q,Lp,uq)\mathcal{D}\left(\mathcal{Q},L^{p},\ell_{u}^{q}\right) (essentially as introduced by Feichtinger and Gr\"obner), where the localized ,,parts`` of a function are measured in the FLp\mathcal{F}L^{p}-norm. This equivalence is useful in several ways: It provides access to a Fourier-analytic understanding of wavelet coorbit spaces, and it allows to discuss coorbit spaces associated to different dilation groups in a common framework. As an illustration of these points, we include a short discussion of dilation invariance properties of coorbit spaces associated to different types of dilation groups

    Approximation in Lp(μ)L^p(\mu) with deep ReLU neural networks

    Full text link
    We discuss the expressive power of neural networks which use the non-smooth ReLU activation function ϱ(x)=max{0,x}\varrho(x) = \max\{0,x\} by analyzing the approximation theoretic properties of such networks. The existing results mainly fall into two categories: approximation using ReLU networks with a fixed depth, or using ReLU networks whose depth increases with the approximation accuracy. After reviewing these findings, we show that the results concerning networks with fixed depth--- which up to now only consider approximation in Lp(λ)L^p(\lambda) for the Lebesgue measure λ\lambda--- can be generalized to approximation in Lp(μ)L^p(\mu), for any finite Borel measure μ\mu. In particular, the generalized results apply in the usual setting of statistical learning theory, where one is interested in approximation in L2(P)L^2(\mathbb{P}), with the probability measure P\mathbb{P} describing the distribution of the data.Comment: Accepted for presentation at SampTA 201

    Design and properties of wave packet smoothness spaces

    Get PDF
    We introduce a family of quasi-Banach spaces - which we call wave packet smoothness spaces - that includes those function spaces which can be characterised by the sparsity of their expansions in Gabor frames, wave atoms, and many other frame constructions. We construct Banach frames for and atomic decompositions of the wave packet smoothness spaces and study their embeddings in each other and in a few more classical function spaces such as Besov and Sobolev spaces.Comment: accepted for publication in Journal de Math\'ematiques Pures et Appliqu\'ee

    From Frazier-Jawerth characterizations of Besov spaces to Wavelets and Decomposition spaces

    Full text link
    This article describes how the ideas promoted by the fundamental papers published by M. Frazier and B. Jawerth in the eighties have influenced subsequent developments related to the theory of atomic decompositions and Banach frames for function spaces such as the modulation spaces and Besov-Triebel-Lizorkin spaces. Both of these classes of spaces arise as special cases of two different, general constructions of function spaces: coorbit spaces and decomposition spaces. Coorbit spaces are defined by imposing certain decay conditions on the so-called voice transform of the function/distribution under consideration. As a concrete example, one might think of the wavelet transform, leading to the theory of Besov-Triebel-Lizorkin spaces. Decomposition spaces, on the other hand, are defined using certain decompositions in the Fourier domain. For Besov-Triebel-Lizorkin spaces, one uses a dyadic decomposition, while a uniform decomposition yields modulation spaces. Only recently, the second author has established a fruitful connection between modern variants of wavelet theory with respect to general dilation groups (which can be treated in the context of coorbit theory) and a particular family of decomposition spaces. In this way, optimal inclusion results and invariance properties for a variety of smoothness spaces can be established. We will present an outline of these connections and comment on the basic results arising in this context

    The universal approximation theorem for complex-valued neural networks

    Full text link
    We generalize the classical universal approximation theorem for neural networks to the case of complex-valued neural networks. Precisely, we consider feedforward networks with a complex activation function σ:CC\sigma : \mathbb{C} \to \mathbb{C} in which each neuron performs the operation CNC,zσ(b+wTz)\mathbb{C}^N \to \mathbb{C}, z \mapsto \sigma(b + w^T z) with weights wCNw \in \mathbb{C}^N and a bias bCb \in \mathbb{C}, and with σ\sigma applied componentwise. We completely characterize those activation functions σ\sigma for which the associated complex networks have the universal approximation property, meaning that they can uniformly approximate any continuous function on any compact subset of Cd\mathbb{C}^d arbitrarily well. Unlike the classical case of real networks, the set of "good activation functions" which give rise to networks with the universal approximation property differs significantly depending on whether one considers deep networks or shallow networks: For deep networks with at least two hidden layers, the universal approximation property holds as long as σ\sigma is neither a polynomial, a holomorphic function, or an antiholomorphic function. Shallow networks, on the other hand, are universal if and only if the real part or the imaginary part of σ\sigma is not a polyharmonic function
    corecore